Goto

Collaborating Authors

 testing strategy


Dynamically Expanding Capacity of Autonomous Driving with Near-Miss Focused Training Framework

Yang, Ziyuan, Li, Zhaoyang, Hu, Jianming, Zhang, Yi

arXiv.org Artificial Intelligence

The long-tail distribution of real driving data poses challenges for training and testing autonomous vehicles (AV), where rare yet crucial safety-critical scenarios are infrequent. And virtual simulation offers a low-cost and efficient solution. This paper proposes a near-miss focused training framework for AV. Utilizing the driving scenario information provided by sensors in the simulator, we design novel reward functions, which enable background vehicles (BV) to generate near-miss scenarios and ensure gradients exist not only in collision-free scenes but also in collision scenarios. And then leveraging the Robust Adversarial Reinforcement Learning (RARL) framework for simultaneous training of AV and BV to gradually enhance AV and BV capabilities, as well as generating near-miss scenarios tailored to different levels of AV capabilities. Results from three testing strategies indicate that the proposed method generates scenarios closer to near-miss, thus enhancing the capabilities of both AVs and BVs throughout training.


Segment Anything Model for Medical Images?

Huang, Yuhao, Yang, Xin, Liu, Lian, Zhou, Han, Chang, Ao, Zhou, Xinrui, Chen, Rusi, Yu, Junxuan, Chen, Jiongquan, Chen, Chaoyu, Liu, Sijing, Chi, Haozhe, Hu, Xindi, Yue, Kejuan, Li, Lei, Grau, Vicente, Fan, Deng-Ping, Dong, Fajin, Ni, Dong

arXiv.org Artificial Intelligence

The Segment Anything Model (SAM) is the first foundation model for general image segmentation. It has achieved impressive results on various natural image segmentation tasks. However, medical image segmentation (MIS) is more challenging because of the complex modalities, fine anatomical structures, uncertain and complex object boundaries, and wide-range object scales. To fully validate SAM's performance on medical data, we collected and sorted 53 open-source datasets and built a large medical segmentation dataset with 18 modalities, 84 objects, 125 object-modality paired targets, 1050K 2D images, and 6033K masks. We comprehensively analyzed different models and strategies on the so-called COSMOS 1050K dataset. Our findings mainly include the following: 1) SAM showed remarkable performance in some specific objects but was unstable, imperfect, or even totally failed in other situations. 2) SAM with the large ViT-H showed better overall performance than that with the small ViT-B. 3) SAM performed better with manual hints, especially box, than the Everything mode. 4) SAM could help human annotation with high labeling quality and less time. 5) SAM was sensitive to the randomness in the center point and tight box prompts, and may suffer from a serious performance drop. 6) SAM performed better than interactive methods with one or a few points, but will be outpaced as the number of points increases. 7) SAM's performance correlated to different factors, including boundary complexity, intensity differences, etc. 8) Finetuning the SAM on specific medical tasks could improve its average DICE performance by 4.39% and 6.68% for ViT-B and ViT-H, respectively. We hope that this comprehensive report can help researchers explore the potential of SAM applications in MIS, and guide how to appropriately use and develop SAM.


Adaptive Sequential Surveillance with Network and Temporal Dependence

Malenica, Ivana, Coyle, Jeremy R., van der Laan, Mark J., Petersen, Maya L.

arXiv.org Machine Learning

Strategic test allocation plays a major role in the control of both emerging and existing pandemics (e.g., COVID-19, HIV). Widespread testing supports effective epidemic control by (1) reducing transmission via identifying cases, and (2) tracking outbreak dynamics to inform targeted interventions. However, infectious disease surveillance presents unique statistical challenges. For instance, the true outcome of interest - one's positive infectious status, is often a latent variable. In addition, presence of both network and temporal dependence reduces the data to a single observation. As testing entire populations regularly is neither efficient nor feasible, standard approaches to testing recommend simple rule-based testing strategies (e.g., symptom based, contact tracing), without taking into account individual risk. In this work, we study an adaptive sequential design involving n individuals over a period of {\tau} time-steps, which allows for unspecified dependence among individuals and across time. Our causal target parameter is the mean latent outcome we would have obtained after one time-step, if, starting at time t given the observed past, we had carried out a stochastic intervention that maximizes the outcome under a resource constraint. We propose an Online Super Learner for adaptive sequential surveillance that learns the optimal choice of tests strategies over time while adapting to the current state of the outbreak. Relying on a series of working models, the proposed method learns across samples, through time, or both: based on the underlying (unknown) structure in the data. We present an identification result for the latent outcome in terms of the observed data, and demonstrate the superior performance of the proposed strategy in a simulation modeling a residential university environment during the COVID-19 pandemic.


Maximizing Software Quality with Artificial Intelligence - RTInsights

#artificialintelligence

Testing solutions that use artificial intelligence help development teams more easily analyze and understand where and what to fix and gives them the ability to more easily analyze and aggregate terabytes of data generated from automated tests. It is no secret that the pandemic has fueled a permanent shift to customer-centric, digital-first experiences, making it essential to provide flawless applications. As such, the field of QA and software testing has become central to building successful development organizations. Innovations such as artificial intelligence (AI) and machine learning solutions that uplevel and automate a number of testing scenarios are becoming necessary to keep up with the growing demand for continuous testing. This includes helping teams to prioritize testing more effectively and only testing new features or pages that are being widely utilized by customers (to save time!).


Machine learning can help slow down future pandemics

#artificialintelligence

In the study, the researchers developed a method to improve testing strategies during epidemic outbreaks and with relatively limited information be able to predict which individuals offer the best potential for testing. "This can be a first step towards society gaining better control of future major outbreaks and reduce the need to shutdown society," says Laura Natali, a doctoral student in physics at the University of Gothenburg and the lead author of the published study. Machine learning is a type of artificial intelligence and can be described as a mathematical model where computers are trained to learn to see connections and solve problems using different data sets. The researchers used machine learning in a simulation of an epidemic outbreak, where information about the first confirmed cases was used to estimate infections in the rest of the population. Data about the infected individual's network of contacts and other information was used: who they have been in close contact with, where and for how long.


A Partially Observable MDP Approach for Sequential Testing for Infectious Diseases such as COVID-19

Singh, Rahul, Liu, Fang, Shroff, Ness B.

arXiv.org Machine Learning

The outbreak of the novel coronavirus (COVID-19) is unfolding as a major international crisis whose influence extends to every aspect of our daily lives. Effective testing allows infected individuals to be quarantined, thus reducing the spread of COVID-19, saving countless lives, and helping to restart the economy safely and securely. Developing a good testing strategy can be greatly aided by contact tracing that provides health care providers information about the whereabouts of infected patients in order to determine whom to test. Countries that have been more successful in corralling the virus typically use a ``test, treat, trace, test'' strategy that begins with testing individuals with symptoms, traces contacts of positively tested individuals via a combinations of patient memory, apps, WiFi, GPS, etc., followed by testing their contacts, and repeating this procedure. The problem is that such strategies are myopic and do not efficiently use the testing resources. This is especially the case with COVID-19, where symptoms may show up several days after the infection (or not at all, there is evidence to suggest that many COVID-19 carriers are asymptotic, but may spread the virus). Such greedy strategies, miss out population areas where the virus may be dormant and flare up in the future. In this paper, we show that the testing problem can be cast as a sequential learning-based resource allocation problem with constraints, where the input to the problem is provided by a time-varying social contact graph obtained through various contact tracing tools. We then develop efficient learning strategies that minimize the number of infected individuals. These strategies are based on policy iteration and look-ahead rules. We investigate fundamental performance bounds, and ensure that our solution is robust to errors in the input graph as well as in the tests themselves.


Revolutionizing API testing with artificial intelligence - SD Times

#artificialintelligence

Recently, a simple conversation I had analyzing the current challenges associated with software testing in the modern era led to a key realization: the tools in the software testing industry have not been focused on simplicity for the Agile world. Agile is primarily a development-focused activity. In its most basic terms, Agile is a software development methodology where typical SDLC activities that would traditionally span over the duration of a project are broken down into much smaller pieces called sprints. Typically, a sprint is 2 to 3 weeks and in a sprint, development activities are focused on new features and enhancements. A sprint starts with the design and creation phase, where the new functionality is split up into user stories, scoped, and then development immediately starts building something.


Revolutionizing API testing with artificial intelligence - SD Times

#artificialintelligence

Recently, a simple conversation I had analyzing the current challenges associated with software testing in the modern era led to a key realization: the tools in the software testing industry have not been focused on simplicity for the Agile world. Agile is primarily a development-focused activity. In its most basic terms, Agile is a software development methodology where typical SDLC activities that would traditionally span over the duration of a project are broken down into much smaller pieces called sprints. Typically, a sprint is 2 to 3 weeks and in a sprint, development activities are focused on new features and enhancements. A sprint starts with the design and creation phase, where the new functionality is split up into user stories, scoped, and then development immediately starts building something.